1,562 research outputs found

    Efficient Coordination in Weakest-Link Games

    Get PDF
    Existing experimental research on behavior in weakest-link games shows overwhelmingly the inability of people to coordinate on the efficient equilibrium, especially in larger groups. We hypothesize that people are able to coordinate on efficient outcomes, provided they have sufficient freedom to choose their interaction neighborhood. We conduct experiments with medium sized and large groups and show that neighborhood choice indeed leads to coordination on the fully efficient equilibrium, irrespective if group size. This leads to substantial welfare effects. Achieved welfare is between 40 and 60 percent higher in games with neighborhood choice than without neighborhood choice. We identify exclusion as the simple but very effective mechanism underlying this result. In early rounds, high performers exclude low performers who in consequence 'learn' to become high performers.efficient coordination, weakest-link, minimum effort, neighborhood choice, experiment

    Efficient Coordination in Weakest-Link Games

    Get PDF
    Existing experimental research on behavior in weakest-link games shows overwhelmingly theinability of people to coordinate on the efficient equilibrium, especially in larger groups. Wehypothesize that people will be able to coordinate on efficient outcomes, provided they havesufficient freedom to choose their interaction neighborhood. We conduct experiments with mediumsized and large groups and show that neighborhood choice indeed leads to coordination on the fullyefficient equilibrium, irrespective of group size. This leads to substantial welfare effects.Achieved welfare is between 40 and 60 percent higher in games with neighborhood choice thanwithout neighborhood choice. We identify exclusion as the simple but very effective mechanismunderlying this result. In early rounds, high performers exclude low performers who in consequence‘learn’ to become high performers.microeconomics ;

    Efficient Coordination in Weakest-Link Games

    Get PDF
    Existing experimental research on behavior in weakest-link games shows overwhelmingly the inability of people to coordinate on the efficient equilibrium, especially in larger groups. We hypothesize that people will be able to coordinate on efficient outcomes, provided they have sufficient freedom to choose their interaction neighborhood. We conduct experiments with medium sized and large groups and show that neighborhood choice indeed leads to coordination on the fully efficient equilibrium, irrespective of group size. This leads to substantial welfare effects. Achieved welfare is between 40 and 60 percent higher in games with neighborhood choice than without neighborhood choice. We identify exclusion as the simple but very effective mechanism underlying this result. In early rounds, high performers exclude low performers who in consequence ‘learn’ to become high performers.efficient coordination, weakest-link, minimum effort, neighborhood choice, experiment

    Controllable Neural Story Plot Generation via Reinforcement Learning

    Full text link
    Language-modeling--based approaches to story plot generation attempt to construct a plot by sampling from a language model (LM) to predict the next character, word, or sentence to add to the story. LM techniques lack the ability to receive guidance from the user to achieve a specific goal, resulting in stories that don't have a clear sense of progression and lack coherence. We present a reward-shaping technique that analyzes a story corpus and produces intermediate rewards that are backpropagated into a pre-trained LM in order to guide the model towards a given goal. Automated evaluations show our technique can create a model that generates story plots which consistently achieve a specified goal. Human-subject studies show that the generated stories have more plausible event ordering than baseline plot generation techniques.Comment: Published in IJCAI 201

    Gender Differences in Performance Under Competition::Is There a Stereotype Threat Shadow?

    Get PDF
    The gender gap in income and leadership positions in many domains of our society is an undisputed pervasive phenomenon. One explanation for the disadvantaged position of women put forward in the economic and psychology literature is the weaker response of women to competitive incentives. Despite the large amount of literature trying to explain this fact, the precise mechanisms behind the gender difference in competitive responsiveness are still not fully uncovered. In this paper, we use laboratory experiments to study the potential role of stereotype threat on the response of men and women to competitive incentives in mixed-gender competition. We use a real effort math task to induce an implicit stereotype threat against women in one treatment. In additional treatments we, respectively, reinforce this stereotype threat and induce a stereotype threat against men. In contrast to much of the literature we do not observe that women are less competitive than men, neither when there is an implicit nor when there is an explicit stereotype threat against women. We attribute this to two factors which differentiates our experiment from previous ones. We control, first, for inter-individual performance differences using a within-subject design, and, second, for risk differences between non-competitive and competitive environments by making the former risky. We do find an adverse stereotype threat effect on the performance of men when there is an explicit stereotype threat against them. In that case any positive performance effect of competition is nullified by the stereotype threat. Overall, our results indicate that a stereotype threat has negative competitive performance effects only if there is information contradicting an existing stereotype. This suggests that the appropriate intervention to prevent the adverse effect of stereotype threat in performance is to avoid any information referring to the stereotype

    Event Representations for Automated Story Generation with Deep Neural Nets

    Full text link
    Automated story generation is the problem of automatically selecting a sequence of events, actions, or words that can be told as a story. We seek to develop a system that can generate stories by learning everything it needs to know from textual story corpora. To date, recurrent neural networks that learn language models at character, word, or sentence levels have had little success generating coherent stories. We explore the question of event representations that provide a mid-level of abstraction between words and sentences in order to retain the semantic information of the original data while minimizing event sparsity. We present a technique for preprocessing textual story data into event sequences. We then present a technique for automated story generation whereby we decompose the problem into the generation of successive events (event2event) and the generation of natural language sentences from events (event2sentence). We give empirical results comparing different event representations and their effects on event successor generation and the translation of events to natural language.Comment: Submitted to AAAI'1

    Unsupervised Methods for Learning and Using Semantics of Natural Language

    Get PDF
    Teaching the computer to understand language is the major goal in the field of natural language processing. In this thesis we introduce computational methods that aim to extract language structure — e.g. grammar, semantics or syntax — from text, which provides the computer with information in order to understand language. During the last decades, scientific efforts and the increase of computational resources made it possible to come closer to the goal of understanding language. In order to extract language structure, many approaches train the computer on manually created resources. Most of these so-called supervised methods show high performance when applied to similar textual data. However, they perform inferior when operating on textual data, which are different to the one they are trained on. Whereas training the computer is essential to obtain reasonable structure from natural language, we want to avoid training the computer using manually created resources. In this thesis, we present so-called unsupervised methods, which are suited to learn patterns in order to extract structure from textual data directly. These patterns are learned with methods that extract the semantics (meanings) of words and phrases. In comparison to manually built knowledge bases, unsupervised methods are more flexible: they can extract structure from text of different languages or text domains (e.g. finance or medical texts), without requiring manually annotated structure. However, learning structure from text often faces sparsity issues. The reason for these phenomena is that in language many words occur only few times. If a word is seen only few times no precise information can be extracted from the text it occurs. Whereas sparsity issues cannot be solved completely, information about most words can be gained by using large amounts of data. In the first chapter, we briefly describe how computers can learn to understand language. Afterwards, we present the main contributions, list the publications this thesis is based on and give an overview of this thesis. Chapter 2 introduces the terminology used in this thesis and gives a background about natural language processing. Then, we characterize the linguistic theory on how humans understand language. Afterwards, we show how the underlying linguistic intuition can be operationalized for computers. Based on this operationalization, we introduce a formalism for representing words and their context. This formalism is used in the following chapters in order to compute similarities between words. In Chapter 3 we give a brief description of methods in the field of computational semantics, which are targeted to compute similarities between words. All these methods have in common that they extract a contextual representation for a word that is generated from text. Then, this representation is used to compute similarities between words. In addition, we also present examples of the word similarities that are computed with these methods. Segmenting text into its topically related units is intuitively performed by humans and helps to extract connections between words in text. We equip the computer with these abilities by introducing a text segmentation algorithm in Chapter 4. This algorithm is based on a statistical topic model, which learns to cluster words into topics solely on the basis of the text. Using the segmentation algorithm, we demonstrate the influence of the parameters provided by the topic model. In addition, our method yields state-of-the-art performances on two datasets. In order to represent the meaning of words, we use context information (e.g. neighboring words), which is utilized to compute similarities. Whereas we described methods for word similarity computations in Chapter 3, we introduce a generic symbolic framework in Chapter 5. As we follow a symbolic approach, we do not represent words using dense numeric vectors but we use symbols (e.g. neighboring words or syntactic dependency parses) directly. Such a representation is readable for humans and is preferred in sensitive applications like the medical domain, where the reason for decisions needs to be provided. This framework enables the processing of arbitrarily large data. Furthermore, it is able to compute the most similar words for all words within a text collection resulting in a distributional thesaurus. We show the influence of various parameters deployed in our framework and examine the impact of different corpora used for computing similarities. Performing computations based on various contextual representations, we obtain the best results when using syntactic dependencies between words within sentences. However, these syntactic dependencies are predicted using a supervised dependency parser, which is trained on language-dependent and human-annotated resources. To avoid such language-specific preprocessing for computing distributional thesauri, we investigate the replacement of language-dependent dependency parsers by language-independent unsupervised parsers in Chapter 6. Evaluating the syntactic dependencies from unsupervised and supervised parses against human-annotated resources reveals that the unsupervised methods are not capable to compete with the supervised ones. In this chapter we use the predicted structure of both types of parses as context representation in order to compute word similarities. Then, we evaluate the quality of the similarities, which provides an extrinsic evaluation setup for both unsupervised and supervised dependency parsers. In an evaluation on English text, similarities computed based on contextual representations generated with unsupervised parsers do not outperform the similarities computed with the context representation extracted from supervised parsers. However, we observe the best results when applying context retrieved by the unsupervised parser for computing distributional thesauri on German language. Furthermore, we demonstrate that our framework is capable to combine different context representations, as we obtain the best performance with a combination of both flavors of syntactic dependencies for both languages. Most languages are not composed of single-worded terms only, but also contain many multi-worded terms that form a unit, called multiword expressions. The identification of multiword expressions is particularly important for semantics, as e.g. the term New York has a different meaning than its single terms New or York. Whereas most research on semantics avoids handling these expressions, we target on the extraction of multiword expressions in Chapter 7. Most previously introduced methods rely on part-of-speech tags and apply a ranking function to rank term sequences according to their multiwordness. Here, we introduce a language-independent and knowledge-free ranking method that uses information from distributional thesauri. Performing evaluations on English and French textual data, our method achieves the best results in comparison to methods from the literature. In Chapter 8 we apply information from distributional thesauri as features for various applications. First, we introduce a general setting for tackling the out-of-vocabulary problem. This problem describes the inferior performance of supervised methods according to words that are not contained in the training data. We alleviate this issue by replacing these unseen words with the most similar ones that are known, extracted from a distributional thesaurus. Using a supervised part-of-speech tagging method, we show substantial improvements in the classification performance for out-of-vocabulary words based on German and English textual data. The second application introduces a system for replacing words within a sentence with a word of the same meaning. For this application, the information from a distributional thesaurus provides the highest-scoring features. In the last application, we introduce an algorithm that is capable to detect the different meanings of a word and groups them into coarse-grained categories, called supersenses. Generating features by means of supersenses and distributional thesauri yields an performance increase when plugged into a supervised system that recognized named entities (e.g. names, organizations or locations). Further directions for using distributional thesauri are presented in Chapter 9. First, we lay out a method, which is capable of incorporating background information (e.g. source of the text collection or sense information) into a distributional thesaurus. Furthermore, we describe an approach on building thesauri for different text domains (e.g. medical or finance domain) and how they can be combined to have a high coverage of domain-specific knowledge as well as a broad background for the open domain. In the last section we characterize yet another method, suited to enrich existing knowledge bases. All three directions might be further extensions, which induce further structure based on textual data. The last chapter gives a summary of this work: we demonstrate that without language-dependent knowledge, a computer can learn to extract useful structure from text by using computational semantics. Due to the unsupervised nature of the introduced methods, we are able to extract new structure from raw textual data. This is important especially for languages, for which less manually created resources are available as well as for special domains e.g. medical or finance. We have demonstrated that our methods achieve state-of-the-art performance. Furthermore, we have proven their impact by applying the extracted structure in three natural language processing tasks. We have also applied the methods to different languages and large amounts of data. Thus, we have not proposed methods, which are suited for extracting structure for a single language, but methods that are capable to explore structure for “language” in general
    corecore